Fast Eigenspace Approximation using Random Signals
نویسندگان
چکیده
We focus in this work on the estimation of the first k eigenvectors of any graph Laplacian using filtering of Gaussian random signals. We prove that we only need k such signals to be able to exactly recover as many of the smallest eigenvectors, regardless of the number of nodes in the graph. In addition, we address key issues in implementing the theoretical concepts in practice using accurate approximated methods. We also propose fast algorithms both for eigenspace approximation and for the determination of the kth smallest eigenvalue λk. The latter proves to be extremely efficient under the assumption of locally uniform distribution of the eigenvalue over the spectrum. Finally, we present experiments which show the validity of our method in practice and compare it to state-of-the-art methods for clustering and visualization both on synthetic smallscale datasets and larger real-world problems of millions of nodes. We show that our method allows a better scaling with the number of nodes than all previous methods while achieving an almost perfect reconstruction of the eigenspace formed by the first k eigenvectors. Keywords—Graph signal processing, low-rank reconstruction, partitionning, spectral graph theory, spectrum analysis, subspace approximation, visualization
منابع مشابه
View-Based Recognition Using an Eigenspace Approximation to the Hausdorff Measure
View based recognition methods such as those using eigenspace techniques have been successful for a number of recognition tasks Currently however such approaches are relatively limited in their ability to recognize objects which are partly hidden from view or occur against cluttered backgrounds In order to address these limitations we have developed a new view matching technique based on an eig...
متن کاملChunk Incremental LDA Computing on Data Streams
This paper presents a constructive method for deriving an updated discriminant eigenspace for classification, when bursts of new classes of data is being added to an initial discriminant eigenspace in the form of random chunks. The proposed Chunk incremental linear discriminant analysis (I-LDA) can effectively evolve a discriminant eigenspace over a fast and large data stream, and extract featu...
متن کاملFast speaker adaptation using eigenspace-based maximum likelihood linear regression
This paper presents an eigenspace-based fast speaker adaptation approach which can improve the modeling accuracy of the conventional maximum likelihood linear regression (MLLR) techniques when only very limited adaptation data is available. The proposed eigenspace-based MLLR approach was developed by introducing a priori knowledge analysis on the training speakers via PCA, so as to construct an...
متن کاملFast Spectral Learning using Lanczos Eigenspace Projections
The core computational step in spectral learning – finding the projection of a function onto the eigenspace of a symmetric operator, such as a graph Laplacian – generally incurs a cubic computational complexity O(N). This paper describes the use of Lanczos eigenspace projections for accelerating spectral projections, which reduces the complexity to O(nTop + nN) operations, where n is the number...
متن کاملRecursive nearest agglomeration (ReNA): fast clustering for approximation of structured signals
In this work, we revisit fast dimension reduction approaches, as with random projections and random sampling. Our goal is to summarize the data to decrease computational costs and memory footprint of subsequent analysis. Such dimension reduction can be very efficient when the signals of interest have a strong structure, such as with images. We focus on this setting and investigate feature clust...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
- CoRR
دوره abs/1611.00938 شماره
صفحات -
تاریخ انتشار 2016